Weak-stability and saddle point theorems for a multiobjective optimization problem with an infinite number of constraints

نویسندگان
چکیده

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Two-Level Optimization Problems with Infinite Number of Convex Lower Level Constraints

‎This paper proposes a new form of optimization problem which is a two-level programming problem with infinitely many lower level constraints‎. ‎Firstly‎, ‎we consider some lower level constraint qualifications (CQs) for this problem‎. ‎Then‎, ‎under these CQs‎, ‎we derive formula for estimating the subdifferential of its valued function‎. ‎Finally‎, ‎we present some necessary optimality condit...

متن کامل

Multiobjective optimization problem with variational inequality constraints

We study a general multiobjective optimization problem with variational inequality, equality, inequality and abstract constraints. Fritz John type necessary optimality conditions involving Mordukhovich coderivatives are derived. They lead to Kuhn-Tucker type necessary optimality conditions under additional constraint qualifications including the calmness condition, the error bound constraint qu...

متن کامل

WEAK AND STRONG DUALITY THEOREMS FOR FUZZY CONIC OPTIMIZATION PROBLEMS

The objective of this paper is to deal with the fuzzy conic program- ming problems. The aim here is to derive weak and strong duality theorems for a general fuzzy conic programming. Toward this end, The convexity-like concept of fuzzy mappings is introduced and then a speci c ordering cone is established based on the parameterized representation of fuzzy numbers. Un- der this setting, duality t...

متن کامل

Generalized iterative methods for solving double saddle point problem

In this paper, we develop some stationary iterative schemes in block forms for solving double saddle point problem. To this end, we first generalize the Jacobi iterative method and study its convergence under certain condition. Moreover, using a relaxation parameter, the weighted version  of the Jacobi method together with its convergence analysis are considered. Furthermore, we extend a method...

متن کامل

On the saddle point problem for non-convex optimization

A central challenge to many fields of science and engineering involves minimizing non-convex error functions over continuous, high dimensional spaces. Gradient descent or quasi-Newton methods are almost ubiquitously used to perform such minimizations, and it is often thought that a main source of difficulty for the ability of these local methods to find the global minimum is the proliferation o...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: TURKISH JOURNAL OF MATHEMATICS

سال: 2019

ISSN: 1303-6149

DOI: 10.3906/mat-1812-16